09. Mini Project: Training an MLP on MNIST

Mini Project: Training an MLP on MNIST

Deep learning is not well-understood, and the practice is ahead of the theory in many cases. If you are new to deep learning, you are strongly encouraged to experiment with many models, to develop intuition about why models work. Starter code is provided on github.

In this mini project, you'll modify the neural network in mnist_mlp.ipynb and compare the resultant model configurations.

Remember: Overfitting is detected by comparing the validation loss to the training loss. If the training loss is much lower than the validation loss, then the model might be overfitting.

Instructions

Task Description:

Take note of the validation loss and test accuracy for the model resulting from each of the below changes. (Try out each amendment in a separate model.)

Task List:

Task Feedback:

Great job! Now, you're ready to learn about Convolutional Neural Networks (CNNs)!

Optional Resource

If you're interested in learning how to do a more systematic hyperparameter search in your neural networks, please check out this blog post.